Outer Alignment
14 pages tagged "Outer Alignment"
Isn’t AI just a tool like any other? Won’t it just do what we tell it to?
Could we tell the AI to do what's morally right?
Can you give an AI a goal which involves “minimally impacting the world”?
At a high level, what is the challenge of AI alignment?
Why can’t we just use Asimov’s Three Laws of Robotics?
Why can't we just make a "child AI" and raise it?
What is the difference between inner and outer alignment?
What is "coherent extrapolated volition (CEV)"?
What is "Do what I mean"?
Which moral theories would be easiest to encode into an AI?
What are "true names" in the context of AI alignment?
What is imitation learning?
What is reward hacking?
What is outer alignment?